Goto

Collaborating Authors

 better training


ReMaX: Relaxing for Better Training on Efficient Panoptic Segmentation

Neural Information Processing Systems

This paper presents a new mechanism to facilitate the training of mask transformers for efficient panoptic segmentation, democratizing its deployment. We observe that due to the high complexity in the training objective of panoptic segmentation, it will inevitably lead to much higher penalization on false positive. Such unbalanced loss makes the training process of the end-to-end mask-transformer based architectures difficult, especially for efficient models. In this paper, we present ReMaX that adds relaxation to mask predictions and class predictions during the training phase for panoptic segmentation. We demonstrate that via these simple relaxation techniques during training, our model can be consistently improved by a clear margin without any extra computational cost on inference.


Better Training of GFlowNets with Local Credit and Incomplete Trajectories

Pan, Ling, Malkin, Nikolay, Zhang, Dinghuai, Bengio, Yoshua

arXiv.org Artificial Intelligence

Generative Flow Networks or GFlowNets are related to Monte-Carlo Markov chain methods (as they sample from a distribution specified by an energy function), reinforcement learning (as they learn a policy to sample composed objects through a sequence of steps), generative models (as they learn to represent and sample from a distribution) and amortized variational methods (as they can be used to learn to approximate and sample from an otherwise intractable posterior, given a prior and a likelihood). They are trained to generate an object $x$ through a sequence of steps with probability proportional to some reward function $R(x)$ (or $\exp(-\mathcal{E}(x))$ with $\mathcal{E}(x)$ denoting the energy function), given at the end of the generative trajectory. Like for other RL settings where the reward is only given at the end, the efficiency of training and credit assignment may suffer when those trajectories are longer. With previous GFlowNet work, no learning was possible from incomplete trajectories (lacking a terminal state and the computation of the associated reward). In this paper, we consider the case where the energy function can be applied not just to terminal states but also to intermediate states. This is for example achieved when the energy function is additive, with terms available along the trajectory. We show how to reparameterize the GFlowNet state flow function to take advantage of the partial reward already accrued at each state. This enables a training objective that can be applied to update parameters even with incomplete trajectories. Even when complete trajectories are available, being able to obtain more localized credit and gradients is found to speed up training convergence, as demonstrated across many simulations.


Better Training using Weight-Constrained Stochastic Dynamics

Leimkuhler, Benedict, Vlaar, Tiffany, Pouchon, Timothée, Storkey, Amos

arXiv.org Machine Learning

We employ constraints to control the parameter space of deep neural networks throughout training. The use of customized, appropriately designed constraints can reduce the vanishing/exploding gradients problem, improve smoothness of classification boundaries, control weight magnitudes and stabilize deep neural networks, and thus enhance the robustness of training algorithms and the generalization capabilities of neural networks. We provide a general approach to efficiently incorporate constraints into a stochastic gradient Langevin framework, allowing enhanced exploration of the loss landscape. We also present specific examples of constrained training methods motivated by orthogonality preservation for weight matrices and explicit weight normalizations. Discretization schemes are provided both for the overdamped formulation of Langevin dynamics and the underdamped form, in which momenta further improve sampling efficiency. These optimization schemes can be used directly, without needing to adapt neural network architecture design choices or to modify the objective with regularization terms, and see performance improvements in classification tasks.


UK Employees will Embrace AI if they get Better Training - UC Today

#artificialintelligence

Genesys, the global leader in omni-channel customer experience and contact centre solutions, recently revealed new research into the state of artificial intelligence (AI) in the UK workplace. According to the research, UK employees are optimistic about the impact that AI could have on their jobs. Around two thirds of the respondents said that they value new tools and technology to help them complete their tasks, and 64% claimed these tools make them more productive at work. Although there are clear benefits in bringing AI into the workforce, there are also steps that companies need to take to drive adoption. Today's team members feel that they need additional support and training to help them make the most of their AI environment.